Your browser doesn't support javascript.
Show: 20 | 50 | 100
Results 1 - 3 de 3
Filter
Add filters

Language
Document Type
Year range
1.
International Journal of Advanced Computer Science and Applications ; 13(10):45-51, 2022.
Article in English | Web of Science | ID: covidwho-2307245

ABSTRACT

Computer programming is a complex field that requires rigorous practice in programming code writing and learning skills, which can be one of the critical challenges in learning and teaching programming. The complicated nature of computer programming requires an instructor to manage its learning resources and diligently generate programming-related questions for students that need conceptual programming and procedural programming rules. In this regard, automatic question generation techniques help teachers carefully align their learning objectives with the question designs in terms of relevancy and complexity. This also helps in reducing the costs linked with the manual generation of questions and fulfills the need of supplying new questions through automatic question techniques. This paper presents a theoretical review of automatic question generation (AQG) techniques, particularly related to computer programming languages from the year 2017 till 2022. A total of 18 papers are included in this study. one of the goals is to analyze and compare the state of the field in question generation before COVID-19 and after the COVID-19 period, and to summarize the challenges and future directions in the field. In congruence to previous studies, there is little focus given in the existing literature on generating questions related to learning programming languages through different techniques. Our findings show that there is a need to further enhance experimental studies in implementing automatic question generation especially in the field of programming. Also, there is a need to implement an authoring tool that can demonstrate designing more practical evaluation metrics for students.

2.
International Journal of Advanced Computer Science and Applications ; 13(10):45-51, 2022.
Article in English | Scopus | ID: covidwho-2145460

ABSTRACT

Computer programming is a complex field that requires rigorous practice in programming code writing and learning skills, which can be one of the critical challenges in learning and teaching programming. The complicated nature of computer programming requires an instructor to manage its learning resources and diligently generate programming-related questions for students that need conceptual programming and procedural programming rules. In this regard, automatic question generation techniques help teachers carefully align their learning objectives with the question designs in terms of relevancy and complexity. This also helps in reducing the costs linked with the manual generation of questions and fulfills the need of supplying new questions through automatic question techniques. This paper presents a theoretical review of automatic question generation (AQG) techniques, particularly related to computer programming languages from the year 2017 till 2022. A total of 18 papers are included in this study. one of the goals is to analyze and compare the state of the field in question generation before COVID-19 and after the COVID-19 period, and to summarize the challenges and future directions in the field. In congruence to previous studies, there is little focus given in the existing literature on generating questions related to learning programming languages through different techniques. Our findings show that there is a need to further enhance experimental studies in implementing automatic question generation especially in the field of programming. Also, there is a need to implement an authoring tool that can demonstrate designing more practical evaluation metrics for students. © 2022, International Journal of Advanced Computer Science and Applications. All Rights Reserved.

3.
Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 ; : 2862-2873, 2021.
Article in English | Scopus | ID: covidwho-1678733

ABSTRACT

The automated transcription of spoken language, and meetings, in particular, is becoming more widespread as automatic speech recognition systems are becoming more accurate. This trend has significantly accelerated since the outbreak of the COVID-19 pandemic, which led to a major increase in the number of online meetings. However, the transcription of spoken language has not received much attention from the NLP community compared to documents and other forms of written language. In this paper, we study a variation of the summarization problem over the transcription of spoken language: given a transcribed meeting, and an action item (i.e., a commitment or request to perform a task), our goal is to generate a coherent and self-contained rephrasing of the action item. To this end, we compiled a novel dataset of annotated meeting transcripts, including human rephrasing of action items. We use state-of-the-art supervised text generation techniques and establish a strong baseline based on BART and UniLM (two pretrained transformer models). Due to the nature of natural speech, language is often broken and incomplete and the task is shown to be harder than an analogous task over email data. Particularly, we show that the baseline models can be greatly improved once models are provided with additional information. We compare two approaches: one incorporating features extracted by coreference-resolution. Additional annotations are used to train an auxiliary model to detect the relevant context in the text. Based on the systematic human evaluation, our best models exhibit near-human-level rephrasing capability on a constrained subset of the problem. © 2021 Association for Computational Linguistics

SELECTION OF CITATIONS
SEARCH DETAIL